# GGUF quantization
Wan2.1 T2V 14B FusionX VACE GGUF
Apache-2.0
This is a text-to-video quantization model that undergoes quantization conversion based on a specific base model and supports various video generation tasks.
Text-to-Video English
W
QuantStack
461
3
Wan2.1 T2V 14B FusionX GGUF
Apache-2.0
This is a quantized text-to-video model that converts the base model to the GGUF format and can be used in ComfyUI, providing more options for text-to-video generation.
Text-to-Video English
W
QuantStack
563
2
Bielik 11B V2.6 Instruct GGUF
Apache-2.0
Bielik-11B-v2.6-Instruct is a large Polish language model developed by SpeakLeash and ACK Cyfronet AGH, fine-tuned based on Bielik-11B-v2, suitable for instruction following tasks.
Large Language Model
Transformers

B
speakleash
206
5
Phantom Wan 1.3B GGUF
Apache-2.0
This is a project that directly converts bytedance-research/Phantom to the GGUF format for image-to-video conversion tasks.
Text-to-Video English
P
QuantStack
315
3
Elisarcyberaiedge7b LoRA GGUF
Apache-2.0
ELISARCyberAIEdge7B-LoRA-GGUF is an offline-ready, quantized LLaMA edge model designed specifically for cybersecurity use cases, effectively addressing various scenarios such as cybersecurity risk assessment.
Large Language Model
Transformers Supports Multiple Languages

E
sallani
786
0
Deepseek R1 0528 GGUF
MIT
A quantized model based on DeepSeek-R1-0528, focusing on text generation tasks and providing a more efficient way of use.
Large Language Model
D
lmstudio-community
1,426
5
Kodify Nano GGUF
Apache-2.0
Kodify-Nano-GGUF is the GGUF version of the Kodify-Nano model, optimized for CPU/GPU inference. It is a lightweight large language model suitable for code development tasks.
Large Language Model
K
MTSAIR
161
1
Voicelab Trurl 2 13b GGUF
trurl-2-13b is a large language model with a parameter scale of 13B, providing a quantized version in GGUF format, suitable for various natural language processing tasks.
Large Language Model Supports Multiple Languages
V
tensorblock
190
1
Bielik 1.5B V3.0 Instruct GGUF
Apache-2.0
This is a 1.5B parameter instruction fine-tuned model for Polish, developed based on the SpeakLeash Bielik series, suitable for text generation tasks.
Large Language Model Other
B
speakleash
341
3
Bielik 4.5B V3.0 Instruct GGUF
Apache-2.0
Bielik-4.5B-v3.0-Instruct-GGUF is a Polish large language model released by SpeakLeash, converted from Bielik-4.5B-v3.0-Instruct to GGUF quantized format, suitable for local inference.
Large Language Model Other
B
speakleash
693
4
Qwen Qwen3 0.6B GGUF
Apache-2.0
This repository contains GGUF format model files for Qwen/Qwen3-0.6B, quantized by TensorBlock's machines and compatible with llama.cpp.
Large Language Model
Q
tensorblock
905
3
Deepcogito Cogito V1 Preview Llama 3B GGUF
A 3B-parameter language model based on the Llama architecture, offering multiple quantization versions to suit different hardware needs
Large Language Model
D
tensorblock
162
1
Kaiden Sakura Violet Square Azura Crimson 12B
Apache-2.0
This is a 12B-parameter large language model based on the fusion of multiple pre-trained language models, merged using the mergekit tool, suitable for text generation tasks.
Large Language Model
Transformers

K
kainatq
81
2
Orpheus 3b Kaya Q8 0.gguf
Apache-2.0
An 8-bit quantized text-to-speech model fine-tuned from Canopy Labs' pre-trained model, supporting 24kHz English audio generation
Speech Synthesis Supports Multiple Languages
O
lex-au
38
0
Gemma 3 1b It GGUF
The GGUF quantized version of the Gemma 3 1B model, suitable for text generation tasks.
Large Language Model
G
MaziyarPanahi
256.05k
4
Qwen2.5 7B VNTL JP EN GGUF
Apache-2.0
GGUF quantized version of the Qwen2.5-7B-based Japanese-English translation model, supporting translation tasks between Japanese and English.
Machine Translation Supports Multiple Languages
Q
TechnoByte
55
1
Qwq 32B GGUF
GGUF format quantized version of QwQ-32B, suitable for local text generation tasks.
Large Language Model
Q
MaziyarPanahi
459.38k
3
Smallthinker 3B Preview Abliterated GGUF
Other
SmallThinker-3B-Preview-abliterated is a 3B-parameter language model based on the GGUF quantized version, suitable for chat and text generation tasks.
Large Language Model English
S
quantflex
122
2
Ltxv Gguf
Other
A GGUF quantized version based on the Lightricks/LTX-Video model, supporting text-to-video, image-to-video, and video-to-video tasks.
Text-to-Video English
L
calcuis
7,378
48
Open Insurance LLM Llama3 8B GGUF
A GGUF quantized version of a specific language model in the insurance field based on NVIDIA Llama 3 - ChatQA, fine-tuned for insurance-related queries and conversations.
Large Language Model
Transformers English

O
Raj-Maharajwala
130
4
Llama3 Aloe 8B Alpha GGUF
Llama3-Aloe-8B-Alpha is an 8B parameter large language model focused on the fields of biology and medicine, offering a quantized version in GGUF format
Large Language Model
Transformers English

L
tensorblock
224
1
Llama 3.2 1B Instruct Q8 0 GGUF
This is Meta's 1 billion parameter instruction-tuned model from the Llama 3.2 series, converted to GGUF format for use with llama.cpp
Large Language Model Supports Multiple Languages
L
hugging-quants
64.04k
31
FLUX.1 Schnell GGUF
Apache-2.0
The .gguf quantized version of FLUX.1-schnell for text-to-image conversion tasks
Text-to-Image
F
OlegSkutte
121
4
Deepseek V2 Lite Chat GGUF
Other
DeepSeek-V2-Lite-Chat is a lightweight chat model optimized based on the DeepSeek-V2 architecture, suitable for efficient dialogue generation tasks.
Large Language Model
Transformers

D
gaianet
1,334
1
Akallama Llama3 70b V0.1 GGUF
Other
AkaLlama is a Korean large language model fine-tuned from Meta-Llama-3-70b-Instruct, focusing on multi-task practical applications
Large Language Model Supports Multiple Languages
A
mirlab
414
15
Mistral 7b Instruct V0.2.q5 K M.gguf
Apache-2.0
Mistral 7B Instruct v0.2 is an instruction fine-tuned version based on the Mistral 7B model, suitable for text generation tasks.
Large Language Model
M
dataequity
114
2
Chatmusician GGUF
MIT
ChatMusician-GGUF is a text generation model based on the GGUF format, suitable for music - related text generation tasks.
Large Language Model
Transformers English

C
MaziyarPanahi
315
13
Lzlv Limarpv3 L2 70b GGUF
This is a static quantization version of the Doctor-Shotgun/lzlv-limarpv3-l2-70b model, offering multiple quantization options to suit different needs.
Large Language Model English
L
mradermacher
67
3
Tinyllama Python Gguf
Apache-2.0
This is the quantized GGUF model file of rahuldshetty/tinyllama-python, fine-tuned based on unsloth/tinyllama-bnb-4bit, specializing in Python code generation tasks.
Large Language Model Supports Multiple Languages
T
rahuldshetty
32
1
Mobilellama 1.4B Base GGUF
Apache-2.0
GGUF quantized version of MobileLLaMA-1.4B-Base, suitable for local deployment and inference
Large Language Model
M
andrijdavid
311
2
Tinyfrank 1.4B
Apache-2.0
Hybrid model based on TinyLlama-1.1B-Chat-v1.0, providing a lightweight LLM solution
Large Language Model
T
SkunkworksAI
120
29
Qcammel 70 X GGUF
Other
qCammel 70 is a large language model based on the Llama 2 architecture, developed by augtoma and quantized by TheBloke. This model focuses on text generation tasks, offering multiple quantization versions to suit different hardware requirements.
Large Language Model English
Q
TheBloke
1,264
4
Featured Recommended AI Models